A Deep Natural Language Inference Predictor Without Language-Specific Training Data
نویسندگان
چکیده
In this paper we present a technique of NLP to tackle the problem inference relation (NLI) between pairs sentences in target language choice without language-specific training dataset. We exploit generic translation dataset, manually translated, along with two instances same pre-trained model - first generate sentence embeddings for source language, and second fine-tuned over mimic first. This is known as Knowledge Distillation. The has been evaluated machine translated Stanford NLI test Multi-Genre RTE3-ITA also proposed architecture different tasks empirically demonstrate generality task. native Italian ABSITA on Sentiment Analysis, Aspect-Based Topic Recognition. emphasise exploitability Distillation that outperforms other methodologies based translation, even though former was not directly trained data it tested over.
منابع مشابه
Annotation Artifacts in Natural Language Inference Data
Large-scale datasets for natural language inference are created by presenting crowd workers with a sentence (premise), and asking them to generate three new sentences (hypotheses) that it entails, contradicts, or is logically neutral with respect to. We show that, in a significant portion of such data, this protocol leaves clues that make it possible to identify the label by looking only at the...
متن کاملNatural Language Inference in Coq
In this paperwe propose away to dealwith natural language inference (NLI) by implementing Modern Type Theoretical Semantics in the proof assistant Coq. The paper is a first attempt to deal with NLI and natural language reasoning in general by using the proof assistant technology. Valid NLIs are treated as theorems and as such the adequacy of our account is tested by trying to prove them. We use...
متن کاملGenerating Natural Language Inference Chains
The ability to reason with natural language is a fundamental prerequisite for many NLP tasks such as information extraction, machine translation and question answering. To quantify this ability, systems are commonly tested whether they can recognize textual entailment, i.e., whether one sentence can be inferred from another one. However, in most NLP applications only single source sentences ins...
متن کاملNatural logic and natural language inference
We propose a model of natural language inference which identifies valid inferences by their lexical and syntactic features, without full semantic interpretation. We extend past work in natural logic, which has focused on semantic containment and monotonicity, by incorporating both semantic exclusion and implicativity. Our model decomposes an inference problem into a sequence of atomic edits lin...
متن کاملTraining a Natural Language Generator From Unaligned Data
We present a novel syntax-based natural language generation system that is trainable from unaligned pairs of input meaning representations and output sentences. It is divided into sentence planning, which incrementally builds deep-syntactic dependency trees, and surface realization. Sentence planner is based on A* search with a perceptron ranker that uses novel differing subtree updates and a s...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Lecture Notes in Computer Science
سال: 2023
ISSN: ['1611-3349', '0302-9743']
DOI: https://doi.org/10.1007/978-3-031-43153-1_15